# Video Self-supervised Learning
Videomaev2 Giant
VideoMAEv2-giant is an ultra-large-scale video classification model based on self-supervised learning, employing a dual masking strategy for pretraining.
Video Processing
V
OpenGVLab
1,071
4
Videomaev2 Large
VideoMAEv2-Large is a large-scale video feature extraction model pre-trained with self-supervision on the UnlabeldHybrid-1M dataset
Video Processing
V
OpenGVLab
5,581
1
Videomae Huge Finetuned Kinetics
VideoMAE is a video pretraining model based on Masked Autoencoder (MAE), fine-tuned on the Kinetics-400 dataset through self-supervised learning, suitable for video classification tasks.
Video Processing
Transformers

V
MCG-NJU
2,984
4
Videomae Base
VideoMAE is a video self-supervised pretraining model based on Masked Autoencoder (MAE), which learns internal video representations by predicting pixel values of masked video patches.
Video Processing
Transformers

V
MCG-NJU
48.66k
45
Videomae Large
VideoMAE is a video self-supervised pre-training model based on Masked Autoencoder (MAE), which learns video representations by predicting pixel values of masked video patches
Video Processing
Transformers

V
MCG-NJU
3,243
31
Videomae Base Short Finetuned Kinetics
VideoMAE is a video self-supervised pre-training model based on Masked Autoencoder (MAE), fine-tuned on the Kinetics-400 dataset for video classification tasks.
Video Processing
Transformers

V
MCG-NJU
62
3
Videomae Base Finetuned Kinetics
VideoMAE is a video self-supervised pre-training model based on Masked Autoencoder (MAE), fine-tuned on the Kinetics-400 dataset for video classification tasks.
Video Processing
Transformers

V
MCG-NJU
44.91k
34
Videomae Base Short
VideoMAE is a video self-supervised pretraining model based on Masked Autoencoder (MAE), which learns internal video representations through masked patch prediction, suitable for downstream tasks like video classification.
Video Processing
Transformers

V
MCG-NJU
886
3
Featured Recommended AI Models